chess position
Exploring Human-AI Conceptual Alignment through the Prism of Chess
Lomasov, Semyon, Goldfeder, Judah, Erol, Mehmet Hamza, So, Matthew, Yan, Yao, Howard, Addison, Kutz, Nathan, Ziv, Ravid Shwartz
Do AI systems truly understand human concepts or merely mimic surface patterns? We investigate this through chess, where human creativity meets precise strategic concepts. Analyzing a 270M-parameter transformer that achieves grandmaster-level play, we uncover a striking paradox: while early layers encode human concepts like center control and knight outposts with up to 85\% accuracy, deeper layers, despite driving superior performance, drift toward alien representations, dropping to 50-65\% accuracy. To test conceptual robustness beyond memorization, we introduce the first Chess960 dataset: 240 expert-annotated positions across 6 strategic concepts. When opening theory is eliminated through randomized starting positions, concept recognition drops 10-20\% across all methods, revealing the model's reliance on memorized patterns rather than abstract understanding. Our layer-wise analysis exposes a fundamental tension in current architectures: the representations that win games diverge from those that align with human thinking. These findings suggest that as AI systems optimize for performance, they develop increasingly alien intelligence, a critical challenge for creative AI applications requiring genuine human-AI collaboration. Dataset and code are available at: https://github.com/slomasov/ChessConceptsLLM.
- North America > Canada > Ontario > Toronto (0.14)
- North America > United States (0.14)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.67)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Games > Chess (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.93)
Explore the Reasoning Capability of LLMs in the Chess Testbed
Wang, Shu, Ji, Lei, Wang, Renxi, Zhao, Wenxiao, Liu, Haokun, Hou, Yifan, Wu, Ying Nian
Reasoning is a central capability of human intelligence. In recent years, with the advent of large-scale datasets, pretrained large language models have emerged with new capabilities, including reasoning. However, these models still struggle with long-term, complex reasoning tasks, such as playing chess. Based on the observation that expert chess players employ a dual approach combining long-term strategic play with short-term tactical play along with language explanation, we propose improving the reasoning capability of large language models in chess by integrating annotated strategy and tactic. Specifically, we collect a dataset named MATE, which consists of 1 million chess positions with candidate moves annotated by chess experts for strategy and tactics. We finetune the LLaMA-3-8B model and compare it against state-of-the-art commercial language models in the task of selecting better chess moves. Our experiments show that our models perform better than GPT, Claude, and Gemini models. We find that language explanations can enhance the reasoning capability of large language models.
- North America > Canada > Ontario > Toronto (0.14)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
Maia-2: A Unified Model for Human-AI Alignment in Chess
Tang, Zhenwei, Jiao, Difan, McIlroy-Young, Reid, Kleinberg, Jon, Sen, Siddhartha, Anderson, Ashton
There are an increasing number of domains in which artificial intelligence (AI) systems both surpass human ability and accurately model human behavior. This introduces the possibility of algorithmically-informed teaching in these domains through more relatable AI partners and deeper insights into human decision-making. Critical to achieving this goal, however, is coherently modeling human behavior at various skill levels. Chess is an ideal model system for conducting research into this kind of human-AI alignment, with its rich history as a pivotal testbed for AI research, mature superhuman AI systems like AlphaZero, and precise measurements of skill via chess rating systems. Previous work in modeling human decision-making in chess uses completely independent models to capture human style at different skill levels, meaning they lack coherence in their ability to adapt to the full spectrum of human improvement and are ultimately limited in their effectiveness as AI partners and teaching tools. In this work, we propose a unified modeling approach for human-AI alignment in chess that coherently captures human style across different skill levels and directly captures how people improve. Recognizing the complex, non-linear nature of human learning, we introduce a skill-aware attention mechanism to dynamically integrate players' strengths with encoded chess positions, enabling our model to be sensitive to evolving player skill. Our experimental results demonstrate that this unified framework significantly enhances the alignment between AI and human players across a diverse range of expertise levels, paving the way for deeper insights into human decision-making and AI-guided teaching tools. Our implementation is available here.
- North America > Canada > Ontario > Toronto (0.28)
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Europe (0.04)
- Research Report > Experimental Study (1.00)
- Research Report > New Finding (0.87)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Natural Language (1.00)
- Information Technology > Artificial Intelligence > Games > Chess (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks (0.93)
Bridging the Human-AI Knowledge Gap: Concept Discovery and Transfer in AlphaZero
Schut, Lisa, Tomasev, Nenad, McGrath, Tom, Hassabis, Demis, Paquet, Ulrich, Kim, Been
Artificial Intelligence (AI) systems have made remarkable progress, attaining super-human performance across various domains. This presents us with an opportunity to further human knowledge and improve human expert performance by leveraging the hidden knowledge encoded within these highly performant AI systems. Yet, this knowledge is often hard to extract, and may be hard to understand or learn from. Here, we show that this is possible by proposing a new method that allows us to extract new chess concepts in AlphaZero, an AI system that mastered the game of chess via self-play without human supervision. Our analysis indicates that AlphaZero may encode knowledge that extends beyond the existing human knowledge, but knowledge that is ultimately not beyond human grasp, and can be successfully learned from. In a human study, we show that these concepts are learnable by top human experts, as four top chess grandmasters show improvements in solving the presented concept prototype positions. This marks an important first milestone in advancing the frontier of human knowledge by leveraging AI; a development that could bear profound implications and help us shape how we interact with AI systems across many AI applications.
- Oceania > Australia > Victoria > Melbourne (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- Europe > Ireland > Leinster > County Dublin > Dublin (0.04)
- Asia > Middle East > Jordan (0.04)
- Information Technology > Artificial Intelligence > Natural Language > Explanation & Argumentation (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.92)
- Information Technology > Artificial Intelligence > Machine Learning > Reinforcement Learning (0.69)
- (3 more...)
The Value of Chess Squares
Gupta, Aditya, Maharaj, Shiva, Polson, Nicholas, Sokolov, Vadim
We propose a neural network-based approach to calculate the value of a chess square-piece combination. Our model takes a triplet (Color, Piece, Square) as an input and calculates a value that measures the advantage/disadvantage of having this piece on this square. Our methods build on recent advances in chess AI, and can accurately assess the worth of positions in a game of chess. The conventional approach assigns fixed values to pieces $(\symking=\infty, \symqueen=9, \symrook=5, \symbishop=3, \symknight=3, \sympawn=1)$. We enhance this analysis by introducing marginal valuations. We use deep Q-learning to estimate the parameters of our model. We demonstrate our method by examining the positioning of Knights and Bishops, and also provide valuable insights into the valuation of pawns. Finally, we conclude by suggesting potential avenues for future research.
- North America > United States > Illinois > Cook County > Chicago (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
Chess2vec: Learning Vector Representations for Chess
Kapicioglu, Berk, Iqbal, Ramiz, Koc, Tarik, Andre, Louis Nicolas, Volz, Katharina Sophia
We conduct the first study of its kind to generate and evaluate vector representations for chess pieces. In particular, we uncover the latent structure of chess pieces and moves, as well as predict chess moves from chess positions. We share preliminary results which anticipate our ongoing work on a neural network architecture that learns these embeddings directly from supervised feedback. The fundamental challenge for machine learning based chess programs is to learn the mapping between chess positions and optimal moves [5, 3, 7]. A chess position is a description of where pieces are located on the chessboard. In learning, chess positions are typically represented as bitboard representations [1]. A bitboard is a 8 8 binary matrix, same dimensions as the chessboard, and each bitboard is associated with a particular piece type (e.g.
- Information Technology > Artificial Intelligence > Machine Learning (1.00)
- Information Technology > Artificial Intelligence > Games > Chess (1.00)
An Algorithm for Automatically Updating a Forsyth-Edwards Notation String Without an Array Board Representation
We present an algorithm that correctly updates the Forsyth-Edwards Notation (FEN) chessboard character string after any move is made without the need for an intermediary array representation of the board. In particular, this relates to software that have to do with chess, certain chess variants and possibly even similar board games with comparable position representation. Even when performance may be equal or inferior to using arrays, the algorithm still provides an accurate and viable alternative to accomplishing the same thing, or when there may be a need for additional or side processing in conjunction with arrays. Furthermore, the end result (i.e. an updated FEN string) is immediately ready for export to any other internal module or external program, unlike with an intermediary array which needs to be first converted into a FEN string for export purposes. The algorithm is especially useful when there are no existing array-based modules to represent a visual board as it can do without them entirely. We provide examples that demonstrate the correctness of the algorithm given a variety of positions involving castling, en passant and pawn promotion.
- Oceania > New Zealand (0.04)
- Europe > United Kingdom > Scotland (0.04)
- Europe > United Kingdom > England > Oxfordshire > Oxford (0.04)
- (5 more...)
DeepChess: End-to-End Deep Neural Network for Automatic Learning in Chess
David, Eli, Netanyahu, Nathan S., Wolf, Lior
We present an end-to-end learning method for chess, relying on deep neural networks. Without any a priori knowledge, in particular without any knowledge regarding the rules of chess, a deep neural network is trained using a combination of unsupervised pretraining and supervised training. The unsupervised training extracts high level features from a given position, and the supervised training learns to compare two chess positions and select the more favorable one. The training relies entirely on datasets of several million chess games, and no further domain specific knowledge is incorporated. The experiments show that the resulting neural network (referred to as DeepChess) is on a par with state-of-the-art chess playing programs, which have been developed through many years of manual feature selection and tuning. DeepChess is the first end-to-end machine learning-based method that results in a grandmaster-level chess playing performance.
- North America > United States > Maryland > Prince George's County > College Park (0.14)
- Asia > Middle East > Israel > Tel Aviv District > Tel Aviv (0.04)
- Europe > Spain > Catalonia > Barcelona Province > Barcelona (0.04)
- Europe > Netherlands > North Holland > Amsterdam (0.04)
Does this chess problem reveal the key to human consciousness?
Artificial intelligence hasn't taken over the world ... yet. But while humans can still outperform computers on most high-level intelligence tasks, at this point most people would concede the game of chess to the machines. The best chess-playing computer programs can already school just about any average human player, and they've proven capable of beating our grandmasters too. But maybe there's still some hope for us, even when it comes to chess. Scientists with the Penrose Institute have devised a unique chess problem that's fairly simple for humans to solve, but which seems to irreparably stump even the most sophisticated of chess programs.
- Information Technology > Artificial Intelligence > Games > Chess (0.54)
- Information Technology > Artificial Intelligence > Issues > Philosophy (0.40)